Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Kv cache e2e add #1000

Closed
wants to merge 6 commits into from
Closed

Kv cache e2e add #1000

wants to merge 6 commits into from

Conversation

horheynm
Copy link
Collaborator

SUMMARY:
"please provide a brief summary"

TEST PLAN:
"please outline how the changes were tested"

Copy link

👋 Hi! Thank you for contributing to llm-compressor. Please add the ready label when the PR is ready for review.

kylesayrs and others added 5 commits December 19, 2024 23:16
* remove sparseml utilities

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>

* use in model_load

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>

* remove use of RECIPE FILE NAME

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>

* rename to RECIPE_FILE_NAME, avoid circular import

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>

* remove qa ignore

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>

* replace tokenizer with processor

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>

* defer data collator changes

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>

---------

Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>
Co-authored-by: Dipika Sikka <dipikasikka1@gmail.com>
* fix offload

Signed-off-by: Dipika <dipikasikka1@gmail.com>

* fix smoothquant offload bug

* remove logtime

---------

Signed-off-by: Dipika <dipikasikka1@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants